Goto

Collaborating Authors

 layer ensemble


Layer Ensemble Averaging for Improving Memristor-Based Artificial Neural Network Performance

Yousuf, Osama, Hoskins, Brian, Ramu, Karthick, Fream, Mitchell, Borders, William A., Madhavan, Advait, Daniels, Matthew W., Dienstfrey, Andrew, McClelland, Jabez J., Lueker-Boden, Martin, Adam, Gina C.

arXiv.org Artificial Intelligence

Artificial neural networks have advanced due to scaling dimensions, but conventional computing faces inefficiency due to the von Neumann bottleneck. This work proposes and experimentally demonstrates layer ensemble averaging - a technique to map pre-trained neural network solutions from software to defective hardware crossbars of emerging memory devices and reliably attain near-software performance on inference. The approach is investigated using a custom 20,000-device hardware prototyping platform on a continual learning problem where a network must learn new tasks without catastrophically forgetting previously learned information. Results demonstrate that by trading off the number of devices required for layer mapping, layer ensemble averaging can reliably boost defective memristive network performance up to the software baseline. For the investigated problem, the average multi-task classification accuracy improves from 61 % to 72 % (< 1 % of software baseline) using the proposed approach. Introduction The increasing demand for large-scale neural network models has prompted a focused exploration of approaches to optimize model efficiency and accelerate computations. Quantized neural networks, which employ reduced-precision representations for model parameters and activations, have emerged as a promising avenue for achieving significant computational gains without compromising performance. As the community delves into extreme quantization, another frontier in enhancing neural network efficiency unfolds through the exploration of emerging memory-based hardware accelerators. For these reasons, memristor-based neural network accelerators have the potential to transform capabilities of artificial intelligence and machine learning systems and thereby usher in a new neuromorphic era of intelligent edge computing. A comprehensive exploration of the interplay between quantized neural networks, dedicated hardware accelerators, and memristive technologies becomes imperative for advancing the capabilities of modern neural network workloads, with the overarching goal of unlocking unprecedented efficiency gains in real-world deep learning applications.


Layer Ensembles

Oleksiienko, Illia, Iosifidis, Alexandros

arXiv.org Artificial Intelligence

Deep Ensembles, as a type of Bayesian Neural Networks, can be used to estimate uncertainty on the prediction of multiple neural networks by collecting votes from each network and computing the difference in those predictions. In this paper, we introduce a method for uncertainty estimation that considers a set of independent categorical distributions for each layer of the network, giving many more possible samples with overlapped layers than in the regular Deep Ensembles. We further introduce an optimized inference procedure that reuses common layer outputs, achieving up to 19x speed up and reducing memory usage quadratically. We also show that the method can be further improved by ranking samples, resulting in models that require less memory and time to run while achieving higher uncertainty quality than Deep Ensembles.